Goto

Collaborating Authors

 main document



Transfer Learning for Classification under Decision Rule Drift with Application to Optimal Individualized Treatment Rule Estimation

Wang, Xiaohan, Ning, Yang

arXiv.org Machine Learning

In this paper, we extend the transfer learning classification framework from regression function-based methods to decision rules. We propose a novel methodology for modeling posterior drift through Bayes decision rules. By exploiting the geometric transformation of the Bayes decision boundary, our method reformulates the problem as a low-dimensional empirical risk minimization problem. Under mild regularity conditions, we establish the consistency of our estimators and derive the risk bounds. Moreover, we illustrate the broad applicability of our method by adapting it to the estimation of optimal individualized treatment rules. Extensive simulation studies and analyses of real-world data further demonstrate both superior performance and robustness of our approach.


# 2): We agree that we should have provided more

Neural Information Processing Systems

W e thank all the three reviewers for their constructive feedback. Please find our answers to major questions raised. Other points will be dealt with in the revised version. Code will be made available by the camera-ready deadline. This discussion will be added to the main document to improve its clarity.


e7feb9dbd9a94b6c552fc403fcebf2ef-Supplemental-Conference.pdf

Neural Information Processing Systems

Organization We provide in-depth descriptions for our algorithms, experimental setups, i.e. dataset configurations, implementation & training details, and additional experimental results & analysis that Section B: We describe dataset configurations for label-and domain-heterogenous scenarios. Section C: We elaborate on implementation and training details for our methods and the baselines. Section D: We provide additional experimental results and analysis. In this section, we describe detailed configurations for datasets that we used in label-and domain-heterogeneous scenarios. These permutations are randomly generated based on different seeds.



Export Reviews, Discussions, Author Feedback and Meta-Reviews

Neural Information Processing Systems

Cramming a full journal paper into the NIPS format is hard, but I would encourage the authors to think of the reader who might want a little more formal vs. descriptive exposition. That is, there is no insight into the "novel transformation" mentioned before Section 3. It would be nice to see some of that transformation described in the main document.


Exploring Prompting Large Language Models as Explainable Metrics

Mahmoudi, Ghazaleh

arXiv.org Artificial Intelligence

This paper describes the IUST NLP Lab submission to the Prompting Large Language Models as Explainable Metrics Shared Task at the Eval4NLP 2023 Workshop on Evaluation & Comparison of NLP Systems. We have proposed a zero-shot prompt-based strategy for explainable evaluation of the summarization task using Large Language Models (LLMs). The conducted experiments demonstrate the promising potential of LLMs as evaluation metrics in Natural Language Processing (NLP), particularly in the field of summarization. Both few-shot and zero-shot approaches are employed in these experiments. The performance of our best provided prompts achieved a Kendall correlation of 0.477 with human evaluations in the text summarization task on the test data. Code and results are publicly available on GitHub.


Fair Clustering via Hierarchical Fair-Dirichlet Process

Chakraborty, Abhisek, Bhattacharya, Anirban, Pati, Debdeep

arXiv.org Artificial Intelligence

The advent of ML-driven decision-making and policy formation has led to an increasing focus on algorithmic fairness. As clustering is one of the most commonly used unsupervised machine learning approaches, there has naturally been a proliferation of literature on {\em fair clustering}. A popular notion of fairness in clustering mandates the clusters to be {\em balanced}, i.e., each level of a protected attribute must be approximately equally represented in each cluster. Building upon the original framework, this literature has rapidly expanded in various aspects. In this article, we offer a novel model-based formulation of fair clustering, complementing the existing literature which is almost exclusively based on optimizing appropriate objective functions.


Convolutional Dictionary Learning

Garcia-Cardona, Cristina, Wohlberg, Brendt

arXiv.org Machine Learning

Convolutional sparse representations are a form of sparse representation with a dictionary that has a structure that is equivalent to convolution with a set of linear filters. While effective algorithms have recently been developed for the convolutional sparse coding problem, the corresponding dictionary learning problem is substantially more challenging. Furthermore, although a number of different approaches have been proposed, the absence of thorough comparisons between them makes it difficult to determine which of them represents the current state of the art. The present work both addresses this deficiency and proposes some new approaches that outperform existing ones in certain contexts. A thorough set of performance comparisons indicates a very wide range of performance differences among the existing and proposed methods, and clearly identifies those that are the most effective.


Probabilistic Backpropagation for Scalable Learning of Bayesian Neural Networks

Hernández-Lobato, José Miguel, Adams, Ryan P.

arXiv.org Machine Learning

Large multilayer neural networks trained with backpropagation have recently achieved state-of-the-art results in a wide range of problems. However, using backprop for neural net learning still has some disadvantages, e.g., having to tune a large number of hyperparameters to the data, lack of calibrated probabilistic predictions, and a tendency to overfit the training data. In principle, the Bayesian approach to learning neural networks does not have these problems. However, existing Bayesian techniques lack scalability to large dataset and network sizes. In this work we present a novel scalable method for learning Bayesian neural networks, called probabilistic backpropagation (PBP). Similar to classical backpropagation, PBP works by computing a forward propagation of probabilities through the network and then doing a backward computation of gradients. A series of experiments on ten real-world datasets show that PBP is significantly faster than other techniques, while offering competitive predictive abilities. Our experiments also show that PBP provides accurate estimates of the posterior variance on the network weights.